Safe Learning: bridging the gap between Bayes, MDL and statistical learning theory via empirical convexity
نویسندگان
چکیده
We extend Bayesian MAP and Minimum Description Length (MDL) learning by testing whether the data can be substantially more compressed by a mixture of the MDL/MAP distribution with another element of the model, and adjusting the learning rate if this is the case. While standard Bayes and MDL can fail to converge if the model is wrong, the resulting “safe” estimator continues to achieve good rates with wrong models. Moreover, when applied to classification and regression models as considered in statistical learning theory, the approach achieves optimal rates under, e.g., Tsybakov’s conditions, and reveals new situations in which we can penalize by (− log prior)/n rather than √ (− log prior)/n.
منابع مشابه
Advanced Course in Machine Learning
The problem of characterizing learnability is the most basic question of learning theory. A fundamental and long-standing answer, formally proven for supervised classification and regression, is that learnability is equivalent to uniform convergence, and that if a problem is learnable, it is learnable via empirical risk minimization. Furthermore, for the problem of binary classification, unifor...
متن کاملReinforcement Learning in Neural Networks: A Survey
In recent years, researches on reinforcement learning (RL) have focused on bridging the gap between adaptive optimal control and bio-inspired learning techniques. Neural network reinforcement learning (NNRL) is among the most popular algorithms in the RL framework. The advantage of using neural networks enables the RL to search for optimal policies more efficiently in several real-life applicat...
متن کاملStatistical Analysis of Regularization Constant - From Bayes, MDL and NIC Points of View
In order to avoid over tting in neural learning, a regularization term is added to the loss function to be minimized. It is naturally derived from the Bayesian standpoint. The present paper studies how to determine the regularization constant from the points of view of the empirical Bayes approach, the maximum description length (MDL) approach, and the network information criterion (NIC) approa...
متن کاملReinforcement Learning in Neural Networks: A Survey
In recent years, researches on reinforcement learning (RL) have focused on bridging the gap between adaptive optimal control and bio-inspired learning techniques. Neural network reinforcement learning (NNRL) is among the most popular algorithms in the RL framework. The advantage of using neural networks enables the RL to search for optimal policies more efficiently in several real-life applicat...
متن کاملActive Learning: An Approach for Reducing Theory-Practice Gap in Clinical Education
Introduction: The gap between theory and practice in clinical fields, including nursing, is one of the main problems that many solutions have been suggested to eliminate it. In this article, we have tried to investigate its solution through active learning. Methods: In this review article, searching articles published during 2000-2012 was done through library references, scientific databases. ...
متن کامل